New Probabilistic Bounds on Eigenvalues and Eigenvectors of Random Kernel Matrices

نویسندگان

  • Nima Reyhani
  • Hideitsu Hino
  • Ricardo Vigário
چکیده

Kernel methods are successful approaches for different machine learning problems. This success is mainly rooted in using feature maps and kernel matrices. Some methods rely on the eigenvalues/eigenvectors of the kernel matrix, while for other methods the spectral information can be used to estimate the excess risk. An important question remains on how close the sample eigenvalues/eigenvectors are to the population values. In this paper, we improve earlier results on concentration bounds for eigenvalues of general kernel matrices. For distance and inner product kernel functions, e.g. radial basis functions, we provide new concentration bounds, which are characterized by the eigenvalues of the sample covariance matrix. Meanwhile, the obstacles for sharper bounds are accounted for and partially addressed. As a case study, we derive a concentration inequality for sample kernel target-alignment.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

APPLICATION OF THE RANDOM MATRIX THEORY ON THE CROSS-CORRELATION OF STOCK ‎PRICES

The analysis of cross-correlations is extensively applied for understanding of interconnections in stock markets. Variety of methods are used in order to search stock cross-correlations including the Random Matrix Theory (RMT), the Principal Component Analysis (PCA) and the Hierachical ‎Structures.‎ In ‎this work‎, we analyze cross-crrelations between price fluctuations of 20 ‎company ‎stocks‎...

متن کامل

On the eigenvalues of some matrices based on vertex degree

The aim of this paper is to compute some bounds of forgotten index and then we present spectral properties of this index. In continuing, we define a new version of energy namely ISI energy corresponded to the ISI index and then we determine some bounds for it.

متن کامل

Asymptotic error bounds for kernel-based Nystrm low-rank approximation matrices

• Many kernel-based learning algorithms have the computational load. • The Nyström low-rank approximation is designed for reducing the computation. • We propose the spectrum decomposition condition with a theoretical justification. • Asymptotic error bounds on eigenvalues and eigenvectors are derived. • Numerical experiments are provided for covariance kernel and Wishart matrix. AMS subject cla...

متن کامل

A note on positive deniteness and stability of interval matrices

It is proved that by using bounds of eigenvalues of an interval matrix, someconditions for checking positive deniteness and stability of interval matricescan be presented. These conditions have been proved previously with variousmethods and now we provide some new proofs for them with a unity method.Furthermore we introduce a new necessary and sucient condition for checkingstability of interval...

متن کامل

Random matrix approach to cross correlations in financial data.

We analyze cross correlations between price fluctuations of different stocks using methods of random matrix theory (RMT). Using two large databases, we calculate cross-correlation matrices C of returns constructed from (i) 30-min returns of 1000 US stocks for the 2-yr period 1994-1995, (ii) 30-min returns of 881 US stocks for the 2-yr period 1996-1997, and (iii) 1-day returns of 422 US stocks f...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2011